HDDS-14921. Improve space accounting in SCM with In-Flight container allocation tracking.#10000
HDDS-14921. Improve space accounting in SCM with In-Flight container allocation tracking.#10000ashishkumar50 wants to merge 7 commits intoapache:masterfrom
Conversation
rakeshadr
left a comment
There was a problem hiding this comment.
Thanks @ashishkumar50 for providing the patch. Added a few comments, please take care.
| if (!alreadyOnDn && getContainerManager() instanceof ContainerManagerImpl) { | ||
| ((ContainerManagerImpl) getContainerManager()) | ||
| .getPendingContainerTracker() | ||
| .removePendingAllocation(dd, id); |
There was a problem hiding this comment.
Say, DN is healthy, all containers confirmed, no new allocations → that DN's bucket never rolls even though heartbeats come every 30 seconds, right?
t=0 Container C1 allocated → pending recorded in tracker
t=60-120 FCR arrives from DN
→ cid = C1
→ alreadyInDn = expectedContainersInDatanode.remove(C1) = FALSE
→ !alreadyInDn = TRUE → removePendingAllocation called → rollIfNeeded fires ✓
→ C1 added to NM DN-set
How abt rolls on every processHeartbeat, every 30 seconds regardless of container state changes ?
| } | ||
|
|
||
| // Cleanup empty buckets to prevent memory leak | ||
| if (bucket.isEmpty()) { |
There was a problem hiding this comment.
Potentially hits concurrency issue. Say two threads entered this block.
Thread-1 (removePendingAllocation): bucket.isEmpty(), returns true
Thread-2 (recordPendingAllocationForDatanode): computeIfAbsent(uuid) returns same bucket
reference (key still exists), calls bucket.add(containerID) and now the bucket will be non-empty
Thread-1: datanodeBuckets.remove(uuid, bucket), then removes the non-empty bucket and now the containerID will be in a detached bucket object, right?
I think, we need to add synchronization to avoid detached bucket object.
| } | ||
|
|
||
| @Test | ||
| public void testRemoveFromBothWindows() { |
There was a problem hiding this comment.
Do we have test scenario covering roll over?
The two-window rolling behavior (container in previousWindow roll after 2× interval). Say, add C1 in currentWindow, then moves C1 to previousWindow, then wait for the roll over.
sumitagrawl
left a comment
There was a problem hiding this comment.
@ashishkumar50 Thanks for working over this, have few review comments.
| ) | ||
| private int transactionToDNsCommitMapLimit = 5000000; | ||
|
|
||
| @Config(key = "hdds.scm.container.pending-allocation.roll-interval", |
There was a problem hiding this comment.
we can have config as hdds.scm.container.pending.allocation.roll.interval
| * @param pipeline The pipeline where container is allocated | ||
| * @param containerID The container being allocated | ||
| */ | ||
| public void recordPendingAllocation(Pipeline pipeline, ContainerID containerID) { |
There was a problem hiding this comment.
This needs to be part of SCMNodeManager, more specific to SCMNodeStat. Reason,
- need handle even like stale node / dead node handler as cleanup
- May need report this when reporting to CLI for available space in the DN
- To be used for pipeline allocation policy, where container manager does not come in role
Its datanode space, just trying to identify already allocated space. And needs to be part of committed space at SCM when reporting to CLI, or other breakup.
| final ContainerInfo container; | ||
| try { | ||
| // Check if container is already known to this DN before adding | ||
| boolean alreadyOnDn = false; |
There was a problem hiding this comment.
Do we really need check if container exist? or just remove if exist as single call ?
There was a problem hiding this comment.
Agreed. Instead of checking and then removing. We can just check the pending list and remove it. It will be the same, We can just avoid one extra op.
There was a problem hiding this comment.
Just checked, This inherently does a copy of the containers with a TreeSetgetExisting(id).copyContainers();, We should avoid this at all cost.
| processContainerReplica(dd, container, replicaProto, publisher, detailsForLogging); | ||
|
|
||
| // Remove from pending tracker when container is added to DN | ||
| if (!alreadyOnDn && getContainerManager() instanceof ContainerManagerImpl) { |
There was a problem hiding this comment.
Please check if node report is also send in ICR, this is for reason that node information should be updated with ICR at same time.
| // (1*5GB) + (2*5GB) = 15GB → actually 3 containers | ||
| long totalCapacity = 0L; | ||
| long effectiveAllocatableSpace = 0L; | ||
| for (StorageReportProto report : storageReports) { |
There was a problem hiding this comment.
Instead of calcuating all available and then removing, we can do progressive base, like,
required=pending+newAllocation
for each report
required = required - volumeUsage in roundoff value
if (required <= 0)
return true
But we need to reserve also, can do first add and check, if not present, remove containerId
OR other way,
when DN report storage handling, total consolidate value can also be added to memory to avoid looping on every call.
| bucket.rollIfNeeded(); | ||
|
|
||
| // Each pending container assumes max size | ||
| return (long) bucket.getCount() * maxContainerSize; |
There was a problem hiding this comment.
This is costly operation as first it combine 2 list, and then it performs count.
| * Contains current and previous window sets, plus last roll timestamp. | ||
| */ | ||
| private static class TwoWindowBucket { | ||
| private Set<ContainerID> currentWindow = ConcurrentHashMap.newKeySet(); |
There was a problem hiding this comment.
if we use synchronized in all methods, then set need not be threadsafe.
| "After 2x this interval, allocations that haven't been confirmed via " + | ||
| "container reports will automatically age out. Default is 10 minutes." | ||
| ) | ||
| private Duration pendingContainerAllocationRollInterval = Duration.ofMinutes(10); |
There was a problem hiding this comment.
rolling period is 5 min or 10 min ? mean previous bucket to be there for additional 5 min to capture containerList IN-Progress
| // Remove from pending tracker when container is added to DN | ||
| // This container was just confirmed for the first time on this DN | ||
| // No need to remove on subsequent reports (it's already been removed) | ||
| if (container != null && getContainerManager() instanceof ContainerManagerImpl) { |
There was a problem hiding this comment.
code handling different from ICR and FCR, can be same only.
| return true; | ||
| } catch (Exception e) { | ||
| LOG.warn("Error checking space for pipeline {}", pipeline.getId(), e); | ||
| return true; |
There was a problem hiding this comment.
If we are not sure if we can create container here, Should we still choose this pipeline? Instead of making it generic, we can specify what to do for each exception we might see.
| // Remove from pending tracker when container is added to DN | ||
| // This container was just confirmed for the first time on this DN | ||
| // No need to remove on subsequent reports (it's already been removed) | ||
| if (container != null && getContainerManager() instanceof ContainerManagerImpl) { |
There was a problem hiding this comment.
Why not just add this to the ContainerManager interface? We can avoid these conversions. Is this because Recon uses the same code path and we don't want it to this? For Recon we can just make it a No-Op.
| final ContainerInfo container; | ||
| try { | ||
| // Check if container is already known to this DN before adding | ||
| boolean alreadyOnDn = false; |
There was a problem hiding this comment.
Agreed. Instead of checking and then removing. We can just check the pending list and remove it. It will be the same, We can just avoid one extra op.
| } | ||
|
|
||
| @Override | ||
| public org.apache.hadoop.hdds.scm.node.DatanodeInfo getDatanodeInfo(DatanodeDetails datanodeDetails) { |
There was a problem hiding this comment.
Nit: We can import DatanodeInfo
| * Get count of all pending containers (union). | ||
| */ | ||
| synchronized int getCount() { | ||
| return getAllPending().size(); |
There was a problem hiding this comment.
Can we just return the count of both windows instead of adding them to separate set just to get the count here.
| * Roll the windows: previous = current, current = empty. | ||
| * Called when current time exceeds lastRollTime + rollIntervalMs. | ||
| */ | ||
| synchronized void rollIfNeeded() { |
There was a problem hiding this comment.
Pending allocations can persist beyond 2× roll interval after long idle periods because rollIfNeeded() only rolls once.A single roll doesn’t clear entries older than two windows which can incorrectly block new allocations.
| long effectiveRemaining = effectiveAllocatableSpace - pendingAllocations; | ||
|
|
||
| // Check if there's enough space for a new container | ||
| if (effectiveRemaining < maxContainerSize) { |
There was a problem hiding this comment.
This makes the allocation little aggressive right? Even if we just have 5GB we allocate it. Should we have leave some buffer when allocating a container?
| bucket.rollIfNeeded(); | ||
|
|
||
| boolean added = bucket.add(containerID); | ||
| LOG.info("Recorded pending container {} on DataNode {}. Added={}, Total pending={}", |
There was a problem hiding this comment.
We can change all the LOG to debug. This will become too noisy on the SCM log for every allocation and removal.
| bucket.rollIfNeeded(); | ||
|
|
||
| boolean removed = bucket.remove(containerID); | ||
| LOG.info("Removed pending container {} from DataNode {}. Removed={}, Remaining={}", |
There was a problem hiding this comment.
Same as above for all the LOG in this class.
What changes were proposed in this pull request?
Maintain space accounting during container allocation in SCM. More detail description is in Jira.
What is the link to the Apache JIRA
https://issues.apache.org/jira/browse/HDDS-14921
How was this patch tested?
UT and IT. (More test scenario TBD)